Developing Your R Workflow

Emorie D Beck

November 1, 2019

What Are Data?

Outline

  1. What is workflow?
  2. Building a codebook
  3. Cleaning your data
  4. Visualizing data before analysis

Starting Your Workflow

A good workflow starts by keeping your files organized outside of R. A typical research project typically involves:

  1. Data
  2. Scripts
  3. Results
  4. Manuscript
  5. Experimental Materials (optional)
  6. Preregistration
  7. Papers

You can set this up outside of R, but I’m going to quickly show you how to set up inside R.

data_path <- "~/Desktop/R_workshop_11_01"
dir.create(data_path)
c("data", "scripts", "results", "manuscript", "experimental materials", 
  "preregistration", "papers") %>%
  paste(data_path, ., sep = "/") %>%
  map(., dir.create)

Workspace

When I create an rmarkdown document for my own research projects, I always start by setting up my my workspace. This involves 3 steps:

  1. Packages
  2. Codebook(s)
  3. Data

Below, we will step through each of these separately, setting ourselves up to (hopefully) flawlessly communicate with R and our data.

Packages

Packages seems like the most basic step, but it is actually very important. ALWAYS LOAD YOUR PACKAGES IN A VERY INTENTIONAL ORDER AT THE BEGINNING OF YOUR SCRIPT. Package conflicts suck, so it needs to be shouted.

# load packages
library(psych)
library(plyr)
library(tidyverse)

Codebook

The second step is a codebook. Arguably, this is the first step because you should create the codebook long before you open R and load your data.

In this case, we are going to using some data from the German Socioeconomic Panel Study (GSOEP), which is an ongoing Panel Study in Germany. Note that these data are for teaching purposes only, shared under the license for the Comprehensive SOEP teaching dataset, which I, as a contracted SOEP user, can use for teaching purposes. These data represent select cases from the full data set and should not be used for the purpose of publication. The full data are available for free at https://www.diw.de/en/diw_02.c.222829.en/access_and_ordering.html.

For this tutorial, I created the codebook for you (Download (won’t work in Safari or IE)), and included what I believe are the core columns you may need. Some of these columns may not be particularly helpful for every dataset.

Codebook

Here are my core columns that are based on the original data:

  1. dataset: this column indexes the name of the dataset that you will be pulling the data from. This is important because we will use this info later on (see purrr tutorial) to load and clean specific data files. Even if you don’t have multiple data sets, I believe consistency is more important and suggest using this.
  2. old_name: this column is the name of the variable in the data you are pulling it from. This should be exact. The goal of this column is that it will allow us to select() variables from the original data file and rename them something that is more useful to us.
  3. item_text: this column is the original text that participants saw or a description of the item.
  4. category: broad categories that different variables can be put into. I’m a fan of naming them things like “outcome”, “predictor”, “moderator”, “demographic”, “procedural”, etc. but sometimes use more descriptive labels like “Big 5” to indicate the model from which the measures are derived.
  5. label: label is basically one level lower than category. So if the category is Big 5, the label would be, or example, “A” for Agreeableness, “SWB” for subjective well-being, etc. This column is most important and useful when you have multiple items in a scales, so I’ll typically leave this blank when something is a standalone variable (e.g. sex, single-item scales, etc.).
  6. item_name: This is the lowest level and most descriptive variable. It indicates which item in scale something is. So it may be “kind” for Agreebleness or “sex” for the demographic biological sex variable.
  7. new_name: This is a column that brings together much of the information we’ve already collected. It’s purpose is to be the new name that we will give to the variable that is more useful and descriptive to us. This is a constructed variable that brings together others. I like to make it a combination of “category”, “label”, “item_name”, and year using varying combos of "_" and “.” that we can use later with tidyverse functions. I typically construct this variable in Excel using the CONCATENATE() function, but it could also be done in R. The reason I do it in Excel is that it makes it easier for someone who may be reviewing my codebook.
  8. scale: this column tells you what the scale of the variable is. Is it a numeric variable, a text variable, etc. This is helpful for knowing the plausible range.
  9. recode: sometimes, we want to recode variables for analyses (e.g. for categorical variables with many levels where sample sizes for some levels are too small to actually do anything with it). I use this column to note the kind of recoding I’ll do to a variable for transparency.

Codebook

Here are additional columns that will make our lives easier or are applicable to some but not all data sets:

  1. reverse: this column tells you whether items in a scale need to be reverse coded. I recommend coding this as 1 (leave alone) and -1 (reverse) for reasons that will become clear later.
  2. mini: this column represents the minimum value of scales that are numeric. Leave blank otherwise.
  3. maxi: this column represents the maximumv alue of scales that are numeric. Leave blank otherwise.
  4. year: for longitudinal data, we have several waves of data and the name of the same item across waves is often different, so it’s important to note to which wave an item belongs. You can do this by noting the wave (e.g. 1, 2, 3), but I prefer the actual year the data were collected (e.g. 2005, 2009, etc.)
  5. meta: Some datasets have a meta name, which essentially means a name that variable has across all waves to make it clear which variables are the same. They are not always useful as some data sets have meta names but no great way of extracting variables using them. But they’re still typically useful to include in your codebook regardless.

Codebook

Key Takeaways from Codebook Use

  1. Reproducibility and Transparency
  2. Creating Composites
  3. Keeping Track of Classes of Variables
  4. Easier Reverse-Coding
  5. Easier Recoding
  6. Your future self will thank you.

Below, I will demonstrate each of these.

Codebook

Below, I’ll load in the codebook we will use for this study, which will include all of the above columns.

# set the path
wd <- "https://github.com/emoriebeck/R-tutorials/blob/master/wustl_r_workshops/workflow"

download.file(
  url      = sprintf("%s/data/codebook.csv?raw=true", wd), 
  destfile = sprintf("%s/data/codebook.csv", data_path)
  )

# load the codebook
(codebook <- sprintf("%s/data/codebook.csv", data_path) %>% 
    read_csv(.) %>%
    mutate(old_name = str_to_lower(old_name)))

Data

First, we need to load in the data. We’re going to use three waves of data from the German Socioeconomic Panel Study, which is a longitudinal study of German households that has been conducted since 1984. We’re going to use more recent data from three waves of personality data collected between 2005 and 2013.

Note: we will be using the teaching set of the GSOEP data set. I will not be pulling from the raw files as a result of this. I will also not be mirroring the format that you would usually load the GSOEP from because that is slightly more complicated and somethng we will return to in a later tutorial on purrr (link) after we have more skills. I’ve left that code in the .Rmd for now, but it won’t make a lot of sense right now.

This code below shows how I would read in and rename a wide-format data set using the codebook I created.

# download the file
download.file(
  url      = sprintf("%s/data/workflow_data.csv?raw=true", wd), 
  destfile = sprintf("%s/data/workflow_data.csv", data_path)
  )

old.names <- codebook$old_name # get old column names
new.names <- codebook$new_name # get new column names

(soep <- sprintf("%s/data/workflow_data.csv", data_path) %>% # path to data
  read_csv(.) %>% # read in data
  select(old.names) %>% # select the columns from our codebook
  setNames(new.names)) # rename columns with our new names

Clean Data

Recode Variables


Write code for at least 1 way of removing the negative values from all columns of the soep data frame.

Activity

In the step above, I showed you one way to recode variables. In this case, using mapvalues() from the plyr package. But we just need to remove all the negative values from our variables (signaling missing or invalid data).


Write code for at least 1 way of removing the negative values from all columns of the soep data frame.

# using ifelse()
(soep <- soep %>%
  mutate_all(~ifelse(. < 0, NA, .)))

# using Base R
soep[soep < 0] <- NA

mapvalues()

  1. the variable you are recoding
  2. a vector of initial values from which you want to
  3. recode your variable to using a vector of new values in the same order as the old values
  4. a way to turn off warnings if some levels are not in your data (warn_missing = F)
(soep <- soep %>%
  mutate_all(~as.numeric(mapvalues(., from = seq(-1,-7, -1), # recode negative 
                to = rep(NA, 7), warn_missing = F)))) # values to NA

Reverse-Scoring

Before we can do that, though, we need to restructure the data a bit in order to bring in the reverse coding information from our codebook. We will talk more about what’s happening here in later tutorials on tidyr, so for now, just bear with me.

Reverse-Scoring

(soep_long <- soep %>%
  gather(key = item, value = value, -contains("Procedural"), # change to long format
         -contains("Demographic"), na.rm = T) %>%
  left_join(codebook %>% select(item = new_name, reverse, mini, maxi)) %>% # bring in codebook
  separate(item, c("type", "item"), sep = "__") %>% # separate category
  separate(item, c("item", "year"), sep = "[.]") %>% # seprate year
  separate(item, c("item", "scrap"), sep = "_") %>% # separate scale and item
  mutate(value = as.numeric(value), # change to numeric
         value = ifelse(reverse == -1, 
            reverse.code(-1, value, mini = mini, maxi = maxi), value)))

Create Composites

Now that we have reverse coded our items, we can create composites.

BFI-S

We’ll start with our scale – in this case, the Big 5 from the German translation of the BFI-S.

Here’s the simplest way, which is also the long way because you’d have to do it for each scale in each year, which I don’t recommend.

soep$C.2005 <- with(soep, rowMeans(cbind(`Big 5__C_thorough.2005`, `Big 5__C_lazy.2005`, 
                          `Big 5__C_efficient.2005`), na.rm = T)) 
soep <- soep %>% select(-C.2005) # get rid of added column

Create Composites

BFI-S

But personally, I don’t have a desire to do that 15 times, so we can use our codebook and dplyr to make our lives a whole lot easier.


Try to create composites using the functions in dplyr for each of the Big 5.
(hint: use the soep_long data frame and filter in only Big 5 items)

Create Composites

BFI-S

But personally, I don’t have a desire to do that 15 times, so we can use our codebook and dplyr to make our lives a whole lot easier.


Try to create composites using the functions in dplyr for each of the Big 5.
(hint: use the soep_long data frame and filter in only Big 5 items)

(b5_soep_long <- soep_long %>%
  filter(type == "Big 5") %>% # keep Big 5 variables
  group_by(Procedural__SID, item, year) %>% # group by person, construct, & year
  summarize(value = mean(value, na.rm = T)) %>% # calculate means
  ungroup() %>% # ungroup
  left_join(soep_long %>% # bring demographic info back in 
    select(Procedural__SID, DOB = Demographic__DOB, Sex = Demographic__Sex) %>%
    distinct()))

Create Composites

Life Events

We also want to create a variable that indexes whether our participants experienced any of the life events during the years of interest (2005-2015).

(events_long <- soep_long %>%
  filter(type == "Life Event") %>% # keep only life events
  group_by(Procedural__SID, item) %>% # group by person and event
  summarize(le_value = sum(value, na.rm = T), # sum up whether they experiened the event at all
            le_value = ifelse(le_value > 1, 1, 0)) %>% # if more than once 1, otherwise 0
   ungroup())

Create Combination Data Set

For the analyses I’d want to do with this data set, I’d want to combine every combination of personality traits with every combination of event we have data for each person for. In other words, I’m looking for a data frame that looks like this:

SID year Event le_value Trait p_value


Try to create this data frame in R.
(hint: you’ll need to use the following functions: gather(), spread(), and full_join().)

Create Combination Data Set

For the analyses I’d want to do with this data set, I’d want to combine every combination of personality traits with every combination of event we have data for each person for. In other words, I’m looking for a data frame that looks like this:

SID year Event le_value Trait p_value


Try to create this data frame in R.
(hint: you’ll need to use the following functions: gather(), spread(), and full_join().)

# Create combined data frame  
(combined_long <- b5_soep_long %>% # take Big Five Data 
  spread(item, value) %>% # Spread it wide 
  full_join(events_long %>% rename(Event = item)) %>% # Join it with Events
  filter(!is.na(Event)) %>% # take out people who don't have event data
  gather(key = Trait, value = p_value, A:O)) # change to doubly long

Save Your Data

  write.csv(
    x = combined_long
    , file = sprintf("%s/data/clean_data_%s.csv", data_path, Sys.Date())
    , row.names = F
    )

Descriptives

Big Five Descriptives

There are lots of ways to create great tables of descriptives. My favorite way is using dplyr, but we will save that for a later lesson on creating great APA style tables in R. For now, we’ll use a wonderfully helpful function from the psych package called describe() in conjunction with a small amount of tidyr to reshape the data.

1. Create a wide data frame that has a column for each personality variable in each year.
2. Call the describe() function on this wide data frame.

Big Five Descriptives

1. Create a wide data frame that has a column for each personality variable in each year.
2. Call the describe() function on this wide data frame.

b5_soep_long  %>%
  unite(tmp, item, year, sep = "_") %>% # make new column that joins item and year
  spread(tmp, value) %>% # make wide because that helps describe
  describe(.) %>% # call describe 
  data.frame %>%
  rownames_to_column(var = "V")

Big Five Distributions

b5_soep_long %>%
  ggplot(aes(x = value)) + 
    geom_histogram() +
    facet_grid(year ~ item) +
    theme_classic()

Life Events Descriptives

For count variables, like life events, we need to use something slightly different. We’re typically more interested in counts – in this case, how many people experienced each life event in the 10 years we’re considering?

To do this, we’ll use a little bit of dplyr rather than the base R function table() that is often used for count data. Instead, we’ll use a combination of group_by() and n() to get the counts by group. In the end, we’re left with a nice little table of counts.

Create a data frame of counts for each life event in R.

Life Events Descriptives

For count variables, like life events, we need to use something slightly different. We’re typically more interested in counts – in this case, how many people experienced each life event in the 10 years we’re considering?

To do this, we’ll use a little bit of dplyr rather than the base R function table() that is often used for count data. Instead, we’ll use a combination of group_by() and n() to get the counts by group. In the end, we’re left with a nice little table of counts.

Create a data frame of counts for each life event in R.

events_long %>%
  group_by(item, le_value) %>% 
  tally() %>%
  ungroup() %>%
  spread(le_value, n)

Life Event Plots

events_long %>%
  mutate(le_value = mapvalues(le_value, 0:1, c("No Event", "Event"))) %>%
  ggplot(aes(x = le_value, fill = le_value)) +
    scale_fill_manual(values = c("cornflowerblue", "gray")) +
    geom_bar(color = "black")  +
    facet_wrap(~item, nrow = 2) +
    theme_classic() +
    theme(legend.position = "bottom")

Scale Reliability

When we work with scales, it’s often a good idea to check the internal consistency of your scale. If the scale isn’t performing how it should be, that could critically impact the inferences you make from your data.

To check the internal consistency of our Big 5 scales, we will use the alpha() function from the psych package, which will give us Cronbach’s as well as a number of other indicators of internal consistency.

Here’s the way you may have seen / done this in the past.

alpha.C.2005 <- with(soep, psych::alpha(x = cbind(`Big 5__C_thorough.2005`, 
                                  `Big 5__C_lazy.2005`, `Big 5__C_efficient.2005`)))

Scale Reliability

But again, doing this 15 times would be quite a pain and would open you up to the possibility of a lot of copy and paste errors.

So instead, to do this, I’m going to use a mix of the tidyverse. At first glance, it may seem complex but as you move through other tutorials (particularly the purrr tutorial), it will begin to make much more sense.

# short function to reshape data and run alpha
alpha_fun <- function(df){
  df %>% spread(scrap,value) %>% psych::alpha(.)
}

(alphas <- soep_long %>%
  filter(type == "Big 5") %>% # filter out Big 5
  select(Procedural__SID, item:value) %>% # get rid of extra columns
  group_by(item, year) %>% # group by construct and year
  nest() %>% # nest the data
  mutate(alpha_res = map(data, alpha_fun), # run alpha
         alpha = map(alpha_res, ~.$total[2])) %>% # get the alpha value
  unnest(alpha)) # pull it out of the list column

Zero-Order Correlations

Finally, we often want to look at the zero-order correlation among study variables to make sure they are performing as we think they should.

To run the correlations, we will need to have our data in wide format, so we’re going to do a little bit of reshaping before we do.

Basically, we need to make a wide format data frame of the Big 5 data similarly to how we did when we called describe() previously.

  1. Create a wide data frame of Big 5 data.
  2. Call the cor() function on it.
  3. Round the resulting matrix of correlations to 2 decimal places.

Zero-Order Correlations

  1. Create a wide data frame of Big 5 data.
  2. Call the cor() function on it.
  3. Round the resulting matrix of correlations to 2 decimal places.
b5_soep_long %>%
  unite(tmp, item, year, sep = "_") %>%
  spread(key = tmp, value = value) %>% 
  select(-Procedural__SID) %>%
  cor(., use = "pairwise") %>%
  round(., 2)
##          DOB   Sex A_2005 A_2009 A_2013 C_2005 C_2009 C_2013 E_2005 E_2009
## DOB     1.00  0.00  -0.08  -0.07  -0.06  -0.13  -0.12  -0.14   0.10   0.12
## Sex     0.00  1.00   0.18   0.17   0.18   0.05   0.07   0.09   0.08   0.08
## A_2005 -0.08  0.18   1.00   0.50   0.50   0.32   0.20   0.19   0.10   0.06
## A_2009 -0.07  0.17   0.50   1.00   0.55   0.19   0.28   0.18   0.05   0.08
## A_2013 -0.06  0.18   0.50   0.55   1.00   0.18   0.19   0.29   0.04   0.06
## C_2005 -0.13  0.05   0.32   0.19   0.18   1.00   0.52   0.48   0.19   0.10
## C_2009 -0.12  0.07   0.20   0.28   0.19   0.52   1.00   0.55   0.12   0.16
## C_2013 -0.14  0.09   0.19   0.18   0.29   0.48   0.55   1.00   0.13   0.14
## E_2005  0.10  0.08   0.10   0.05   0.04   0.19   0.12   0.13   1.00   0.61
## E_2009  0.12  0.08   0.06   0.08   0.06   0.10   0.16   0.14   0.61   1.00
## E_2013  0.10  0.11   0.04   0.04   0.07   0.10   0.10   0.18   0.59   0.65
## N_2005  0.06 -0.18   0.10   0.06   0.02   0.09   0.06   0.03   0.18   0.10
## N_2009  0.03 -0.22   0.07   0.09   0.03   0.06   0.08   0.05   0.13   0.16
## N_2013  0.02 -0.21   0.06   0.06   0.10   0.04   0.06   0.08   0.10   0.10
## O_2005  0.11  0.06   0.12   0.09   0.07   0.17   0.12   0.08   0.40   0.29
## O_2009  0.10  0.05   0.05   0.11   0.07   0.06   0.14   0.08   0.26   0.36
## O_2013  0.05  0.07   0.08   0.09   0.13   0.07   0.08   0.15   0.24   0.28
##        E_2013 N_2005 N_2009 N_2013 O_2005 O_2009 O_2013
## DOB      0.10   0.06   0.03   0.02   0.11   0.10   0.05
## Sex      0.11  -0.18  -0.22  -0.21   0.06   0.05   0.07
## A_2005   0.04   0.10   0.07   0.06   0.12   0.05   0.08
## A_2009   0.04   0.06   0.09   0.06   0.09   0.11   0.09
## A_2013   0.07   0.02   0.03   0.10   0.07   0.07   0.13
## C_2005   0.10   0.09   0.06   0.04   0.17   0.06   0.07
## C_2009   0.10   0.06   0.08   0.06   0.12   0.14   0.08
## C_2013   0.18   0.03   0.05   0.08   0.08   0.08   0.15
## E_2005   0.59   0.18   0.13   0.10   0.40   0.26   0.24
## E_2009   0.65   0.10   0.16   0.10   0.29   0.36   0.28
## E_2013   1.00   0.11   0.13   0.15   0.26   0.28   0.35
## N_2005   0.11   1.00   0.55   0.53   0.09   0.08   0.06
## N_2009   0.13   0.55   1.00   0.60   0.06   0.07   0.07
## N_2013   0.15   0.53   0.60   1.00   0.05   0.05   0.05
## O_2005   0.26   0.09   0.06   0.05   1.00   0.58   0.55
## O_2009   0.28   0.08   0.07   0.05   0.58   1.00   0.61
## O_2013   0.35   0.06   0.07   0.05   0.55   0.61   1.00

Zero-Order Correlations Plot

This is a lot of values and a little hard to make sense of, so as a bonus, I’m going to give you a little bit of more complex code that makes this more readable (and publishable ).

r <- b5_soep_long %>%
  unite(tmp, item, year, sep = "_") %>%
  spread(key = tmp, value = value) %>% 
  select(-Procedural__SID, -DOB, -Sex) %>%
  cor(., use = "pairwise") 

r[upper.tri(r, diag = T)] <- NA
diag(r) <- (alphas %>% arrange(item, year))$std.alpha

r %>% data.frame %>%
  rownames_to_column("V1") %>%
  gather(key = V2, value = r, na.rm = T, -V1) %>%
  separate(V1, c("T1", "Year1"), sep = "_") %>%
  separate(V2, c("T2", "Year2"), sep = "_") %>%
  mutate_at(vars(Year1), ~factor(., levels = c(2013, 2009, 2005))) %>%
  ggplot(aes(x = Year2, y = Year1, fill = r)) +
    geom_raster() + 
  scale_fill_gradient2(low = "blue", high = "red", mid = "white", 
   midpoint = 0, limit = c(-1,1), space = "Lab", 
   name="Correlations") +  
  geom_text(aes(label = round(r,2))) +
    facet_grid(T1 ~ T2) +
    theme_classic()
Correlations among Personality Indicators. Values on the diagonal represent Chronbach's alpha for each scale in each year. Within-trait correlations represent test-retest correlations.

Correlations among Personality Indicators. Values on the diagonal represent Chronbach’s alpha for each scale in each year. Within-trait correlations represent test-retest correlations.